Representing words as regions in vector space

نویسنده

  • Katrin Erk
چکیده

Vector space models of word meaning typically represent the meaning of a word as a vector computed by summing over all its corpus occurrences. Words close to this point in space can be assumed to be similar to it in meaning. But how far around this point does the region of similar meaning extend? In this paper we discuss two models that represent word meaning as regions in vector space. Both representations can be computed from traditional point representations in vector space. We find that both models perform at over 95% F-score on a token classification task.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

An image based feature space and mapping for linking regions and words

We propose an image based feature space and define a mapping of both image regions and textual labels into that space. We believe the embedding of both image regions and labels into the same space in this way is novel, and makes object recognition more straightforward. Each dimension of the space corresponds to an image from the database. The coordinates of an image segment(region) are calculat...

متن کامل

A Comparative Study for WordNet Guided Text Representation

Text information processing depends critically on the proper text representation. A common and naïve way of representing a document is a bag of its component words [1], but the semantic relations between words are ignored, such as synonymy and hypernymy-hyponymy between nouns. This paper presents a model for representing a document in terms of the synonymy sets (synsets) in WordNet [2]. The syn...

متن کامل

Word Association Spaces for Predicting Semantic Similarity Effects in Episodic Memory

A common assumption of theories of memory is that the meaning of a word can be represented by a vector which places a word as a point in a multidimensional semantic space (e.g. Landauer & Dumais, 1997; Burgess & Lund, 2000; Osgood, Suci, & Tannenbaum, 1957). Representing words as vectors in a multidimensional space allows simple geometric operations such as the Euclidian distance or the angle b...

متن کامل

Semantic Spaces based on Free Association that Predict Memory Performance

Many memory models represent aspects of words such as meaning by vectors of feature values, such that words with similar meanings are placed in similar regions of the semantic space whose dimensions are defined by the vector positions. Methods for constructing such spaces include those based on scaling similarity ratings for pairs of words, and those based on the analysis of co-occurrence stati...

متن کامل

Cross Lingual Lexical Substitution Using Word Representation in Vector Space

Polysemous words acquire different senses and meanings from their contexts. Representing words in vector space as a function of their contexts captures some semantic and syntactic features for words and introduces new useful relations between them. In this paper, we exploit different vectorized representations for words to solve the problem of Cross Lingual Lexical Substitution. We compare our ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2009